AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
GPU-efficient inference

# GPU-efficient inference

BAGEL 7B MoT DF11
The BAGEL-7B-MoT model utilizing DFloat11 lossless compression technology reduces volume by 32% while maintaining bit-level output consistency
Large Language Model Safetensors
B
DFloat11
428
8
Paraformer Large
Apache-2.0
Paraformer is an innovative non-autoregressive end-to-end speech recognition model with significant advantages over traditional autoregressive models. It can generate entire target sentences in parallel, making it particularly suitable for GPU-accelerated parallel inference.
Speech Recognition Chinese
P
funasr
43
45
Gpt J 6B
GPT-J 6B is a Transformer model based on the GPT-3 architecture, featuring 6 billion parameters and supporting text generation tasks.
Large Language Model Transformers
G
flyhero
59
13
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase